Scalable Bayesian modelling for smoothing disease risks in large spatial data sets using INLA

نویسندگان

چکیده

Several methods have been proposed in the spatial statistics literature to analyse big data sets continuous domains. However, new for analysing high-dimensional areal are still scarce. Here, we propose a scalable Bayesian modelling approach smoothing mortality (or incidence) risks data, that is, when number of small areas is very large. The method implemented R add-on package bigDM and it based on idea “divide conquer“. Although this proposal could possibly be using any fitting technique, use INLA here (integrated nested Laplace approximations) as now well-known computationally efficient, easy practitioners handle. We proposal’s empirical performance comprehensive simulation study considers two model-free settings. Finally, methodology applied male colorectal cancer Spanish municipalities showing its benefits with regard standard terms goodness fit computational time.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Bayesian Spatial Modelling with R-INLA

The principles behind the interface to continuous domain spatial models in the RINLA software package for R are described. The Integrated Nested Laplace Approximation (INLA) approach proposed by Rue, Martino, and Chopin (2009) is a computationally effective alternative to MCMC for Bayesian inference. INLA is designed for latent Gaussian models, a very wide and flexible class of models ranging f...

متن کامل

Scalable Anonymization Algorithms for Large Data Sets

k-Anonymity is a widely-studied mechanism for protecting identity when distributing non-aggregate personal data. This basic mechanism can also be extended to protect an individual-level sensitive attribute. Numerous algorithms have been developed in recent years for generalizing, clustering, or otherwise manipulating data to satisfy one or more anonymity requirements. However, few have consider...

متن کامل

Approximating likelihoods for large spatial data sets

Likelihood methods are often difficult to use with large, irregularly sited spatial data sets, owing to the computational burden. Even for Gaussian models, exact calculations of the likelihood for n observations require O.n3/ operations. Since any joint density can be written as a product of conditional densities based on some ordering of the observations, one way to lessen the computations is ...

متن کامل

A Novel Hyperbolic Smoothing Algorithm for Clustering Large Data Sets

The minimum sum-of-squares clustering problem is considered. The mathematical modeling of this problem leads to a min − sum −min formulation which, in addition to its intrinsic bi-level nature, has the significant characteristic of being strongly nondifferentiable. To overcome these difficulties, the proposed resolution method, called Hyperbolic Smoothing, adopts a smoothing strategy using a sp...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: spatial statistics

سال: 2021

ISSN: ['2211-6753']

DOI: https://doi.org/10.1016/j.spasta.2021.100496